Word Acquisition in Neural Language Models

نویسندگان

چکیده

Abstract We investigate how neural language models acquire individual words during training, extracting learning curves and ages of acquisition for over 600 on the MacArthur-Bates Communicative Development Inventory (Fenson et al., 2007). Drawing studies word in children, we evaluate multiple predictors words’ LSTMs, BERT, GPT-2. find that effects concreteness, length, lexical class are pointedly different children models, reinforcing importance interaction sensorimotor experience child acquisition. Language rely far more frequency than but, like they exhibit slower longer utterances. Interestingly, follow consistent patterns training both unidirectional bidirectional LSTM Transformer architectures. Models predict based unigram token frequencies early before transitioning loosely to bigram probabilities, eventually converging nuanced predictions. These results shed light role distributional mechanisms while also providing insights human-like models.

برای دانلود باید عضویت طلایی داشته باشید

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Word Order Acquisition in Persian Speaking Children

Objectives: Persian is a pro-drop language with canonical Subject-Object-Verb (SOV) word order. This study investigates the acquisition of word order in Persian-speaking children. Methods: In the present study, participants were 60 Persian-speaking children (30 girls and 30 boys) with typically developing language skills, and aged between 30-47 months. The 30-minute language samples were audio...

متن کامل

learners’ attitudes toward the effectiveness of mobile-assisted language learning (mall) in vocabulary acquisition in the iranian efl context: the case of word lists, audiobooks and dictionary use

رشد انفجاری تکنولوژی فرصت های آموزشی مهیج و جدیدی را پیش روی فراگیران و آموزش دهندگان گذاشته است. امروزه معلمان برای اینکه در امر آموزش زبان بروز باشند باید روش هایی را اتخاذ نمایند که درآن ها از تکنولوژی جهت کمک در یادگیری زبان دوم و چندم استفاده شده باشد. با در نظر گرفتن تحولاتی که رشته ی آموزش زبان در حال رخ دادن است هم اکنون زمان مناسبی برای ارزشیابی نگرش های موجود نسبت به تکنولوژی های جدید...

15 صفحه اول

Using Factored Word Representation in Neural Network Language Models

Neural network language and translation models have recently shown their great potentials in improving the performance of phrase-based machine translation. At the same time, word representations using different word factors have been translation quality and are part of many state-of-theart machine translation systems. used in many state-of-the-art machine translation systems, in order to suppor...

متن کامل

Neural Network Models for Language Acquisition: A Brief Survey

Since the outbreak of connectionist modelling in the mid eighties, several problems in natural language processing have been tackled by employing neural network-based techniques. Neural network's biological plausibility o ers a promising framework in which the computational treatment of language may be linked to other disciplines such as cognitive science and psychology. With this brief survey,...

متن کامل

Compressing Neural Language Models by Sparse Word Representations

Neural networks are among the state-ofthe-art techniques for language modeling. Existing neural language models typically map discrete words to distributed, dense vector representations. After information processing of the preceding context words by hidden layers, an output layer estimates the probability of the next word. Such approaches are timeand memory-intensive because of the large number...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

ژورنال

عنوان ژورنال: Transactions of the Association for Computational Linguistics

سال: 2022

ISSN: ['2307-387X']

DOI: https://doi.org/10.1162/tacl_a_00444